Topological Orders Based Planning for Solving POMDPs
نویسندگان
چکیده
Although partially observable Markov decision processes (POMDPs) have received significant attention in past years, to date, solving problems of realistic order of magnitude remains a serious challenge. In this context, techniques that accelerate fundamental algorithms have been a main focus of research. Among them prioritized solvers suggest solutions to the problem of ordering backup operations. Prioritization techniques for ordering the sequence of backup operations considerably reduce the number of backups needed, but involve significant overhead. This paper introduces a novel prioritized method, namely topological order-based planning (TOP), that exploits causal relations between states to deal with two key issues. First, TOP detects the structure of POMDPs as a means of overcoming both the dimensionality and the history curses. Second, it circumvents the problem of unnecessary backups and builds approximate solutions based on a topological order induced by the underlying structure. Empirical experiments prove that TOP is competitive with the best techniques on general domains, and can perform significantly better on layered ones.
منابع مشابه
Exploiting locality of interaction in factored Dec-POMDPs
Decentralized partially observable Markov decision processes (Dec-POMDPs) constitute an expressive framework for multiagent planning under uncertainty, but solving them is provably intractable. We demonstrate how their scalability can be improved by exploiting locality of interaction between agents in a factored representation. Factored Dec-POMDP representations have been proposed before, but o...
متن کاملApproximate Solutions for Factored Dec-POMDPs with Many Agents1
Dec-POMDPs are a powerful framework for planning in multiagent systems, but are provably intractable to solve. This paper proposes a factored forward-sweep policy computation method that tackles the stages of the problem one by one, exploiting weakly coupled structure at each of these stages. An empirical evaluation shows that the loss in solution quality due to these approximations is small an...
متن کاملApproximate Solutions for Factored Dec-POMDPs with Many Agents — Extended Abstract1
Dec-POMDPs are a powerful framework for planning in multiagent systems, but are provably intractable to solve. This paper proposes a factored forward-sweep policy computation method that tackles the stages of the problem one by one, exploiting weakly coupled structure at each of these stages. An empirical evaluation shows that the loss in solution quality due to these approximations is small an...
متن کاملExponential Family PCA for Belief Compression in POMDPs
Standard value function approaches to finding policies for Partially Observable Markov Decision Processes (POMDPs) are intractable for large models. The intractability of these algorithms is due to a great extent to their generating an optimal policy over the entire belief space. However, in real POMDP problems most belief states are unlikely, and there is a structured, low-dimensional manifold...
متن کاملMonte-Carlo Planning in Large POMDPs
This paper introduces a Monte-Carlo algorithm for online planning in large POMDPs. The algorithm combines a Monte-Carlo update of the agent’s belief state with a Monte-Carlo tree search from the current belief state. The new algorithm, POMCP, has two important properties. First, MonteCarlo sampling is used to break the curse of dimensionality both during belief state updates and during planning...
متن کامل